Guiding CLASSY Toward More Responsive Summaries
نویسندگان
چکیده
We discuss changes and improvements in CLASSY for TAC 2010 along with a set of evaluation metrics. Results for both summarization and AESOP evaluation are given.
منابع مشابه
Arabic/English Multi-document Summarization with CLASSY - The Past and the Future
Automatic document summarization has become increasingly important due to the quantity of written material generated worldwide. Generating good quality summaries enables users to cope with larger amounts of information. English-document summarization is a difficult task. Yet it is not sufficient. Environmental, economic, and other global issues make it imperative for English speakers to underst...
متن کاملCLASSY 2009: Summarization and Metrics
This year the CLASSY team participated in the update summary task and made four submissions to summarization evaluation (AESOP). Our AESOP submissions used combinations of ROUGE scores along with an update (or newness) score. We also use these new metrics, which we call Nouveau ROUGE, to help train our system and evaluate new ideas on computing update summaries. CLASSY (Clustering, Linguistics,...
متن کاملCLASSY 2011 at TAC: Guided and Multi-lingual Summaries and Evaluation Metrics
We present CLASSY’s guided summarization as well as multi-lingual methods as submitted to TAC 2011. In addition, we describe improved metrics submitted to the AESOP task at TAC.
متن کاملCLASSY Query-Based Multi-Document Summarization
Our summarizer is based on an HMM (Hidden Markov Model) for sentence selection within a document and a pivoted QR algorithm to generate a multi-document summary. Each year, since we began participating in DUC in 2001, we have modified the features used by the HMM and have added linguistic capabilities in order to improve the summaries we generate. Our system, called “CLASSY” (Clustering, Lingui...
متن کاملAssessing the Effect of Inconsistent Assessors on Summarization Evaluation
We investigate the consistency of human assessors involved in summarization evaluation to understand its effect on system ranking and automatic evaluation techniques. Using Text Analysis Conference data, we measure annotator consistency based on human scoring of summaries for Responsiveness, Readability, and Pyramid scoring. We identify inconsistencies in the data and measure to what extent the...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2010